8 research outputs found

    SALSA: A Novel Dataset for Multimodal Group Behavior Analysis

    Get PDF
    Studying free-standing conversational groups (FCGs) in unstructured social settings (e.g., cocktail party ) is gratifying due to the wealth of information available at the group (mining social networks) and individual (recognizing native behavioral and personality traits) levels. However, analyzing social scenes involving FCGs is also highly challenging due to the difficulty in extracting behavioral cues such as target locations, their speaking activity and head/body pose due to crowdedness and presence of extreme occlusions. To this end, we propose SALSA, a novel dataset facilitating multimodal and Synergetic sociAL Scene Analysis, and make two main contributions to research on automated social interaction analysis: (1) SALSA records social interactions among 18 participants in a natural, indoor environment for over 60 minutes, under the poster presentation and cocktail party contexts presenting difficulties in the form of low-resolution images, lighting variations, numerous occlusions, reverberations and interfering sound sources; (2) To alleviate these problems we facilitate multimodal analysis by recording the social interplay using four static surveillance cameras and sociometric badges worn by each participant, comprising the microphone, accelerometer, bluetooth and infrared sensors. In addition to raw data, we also provide annotations concerning individuals' personality as well as their position, head, body orientation and F-formation information over the entire event duration. Through extensive experiments with state-of-the-art approaches, we show (a) the limitations of current methods and (b) how the recorded multiple cues synergetically aid automatic analysis of social interactions. SALSA is available at http://tev.fbk.eu/salsa.Comment: 14 pages, 11 figure

    Multimodal Personality Recognition from Audiovisual Data

    Get PDF
    Automatic behavior analysis lies at the intersection of diļ¬€erent social and technical research domains. The interdisciplinarity of the ļ¬eld, provides researchers with the means to study the manifestations of human constructs, such as personality. A branch of human behavior analyis, the study of personality provides insight into the cognitive and psychological construction of the human being. Research in personality psychology, advances in com- puting power and the development of algorithms, have made it possible to analyze existing data in order to understand how people express their own personality, perceive othersā€™, and what are the variables that inļ¬‚uence its manifestation. We are pursuing this line of research because insights into the personality of the user can have an impact on how we interact with technology. Incorporating research on personality recogniton, both from a cognitive as well as an engineering perspective, into computers could facilitate the interactions between humans and machines. Previous attempts on personality recognition have focused on a variety of diļ¬€erent corporas (ranging from text to audiovisual data), different scenarios (interviews, meetings), different channels of communication (audio, video, text) and different subsets of personality traits (out of the ļ¬ve ones present in the Big Five Model: Extraversion, Agreeableness, Conscientiousness, Emotional Stability and Creativity). Our work builds on previous research, by considering simple acoustic and visual non-verbal features extracted from multimodal data, but doesnā€™t fail to bring novelties: we consider previously uninvestigated scenarios, and at the same time, all of the ļ¬ve personality traits and not just a subset

    Multimodal Personality Recognition in Collaborative Goal-Oriented Tasks

    No full text
    Incorporating research on personality recognition into computers, both from a cognitive as well as an engineering perspective, would facilitate the interactions between humans and machines. Previous attempts on personality recognition have focused on a variety of different corpora (ranging from text to audiovisual data), scenarios (interviews, meetings), channels of communication (audio, video, text), and different subsets of personality traits (out of the five ones from the Big Five Model). Our study uses simple acoustic and visual nonverbal features extracted from multimodal data, which have been recorded in previously uninvestigated scenarios, and consider all five personality traits and not just a subset. First, we look at the human-machine interaction scenario, where we introduce the display of different ā€œcollaboration levels.ā€ Second, we look at the contribution of the human-human interaction (HHI) scenario on the emergence of personality traits. Investigating the HHI scenario creates a stronger basis for future human-agents interactions. Our goal is to study, from a computational approach, the emergence degree of the five personality traits in these two scenarios. The results demonstrate the relevance of each of the two scenarios when it comes to the degree of emergence of certain traits and the feasibility to automatically recognize personality under different conditions

    Multimodal Personality Recognition in Collaborative Goal-Oriented Tasks

    No full text

    SALSA: A Novel Dataset for Multimodal Group Behaviour Analysis

    No full text
    Studying free-standing conversational groups (FCGs) in unstructured social settings (e.g., cocktail party) is gratifying due to the wealth of information available at the group (mining social networks) and individual (recognizing native behavioral and personality traits) levels. However, analyzing social scenes involving FCGs is also highly challenging due to the difficulty in extracting behavioral cues such as target locations, their speaking activity and head/body pose due to crowdedness and presence of extreme occlusions. To this end, we propose SALSA, a novel dataset facilitating multimodal and Synergetic sociAL Scene Analysis, and make two main contributions to research on automated social interaction analysis: (1) SALSA records social interactions among 18 participants in a natural, indoor environment for over 60 minutes, under the poster presentation and cocktail party contexts presenting difficulties in the form of low-resolution images, lighting variations, numerous occlusions, reverberations and interfering sound sources; (2) To alleviate these problems we facilitate multimodal analysis by recording the social interplay using four static surveillance cameras and sociometric badges worn by each participant, comprising the microphone, accelerometer, bluetooth and infrared sensors. In addition to raw data, we also provide annotations concerning individualsā€™ personality as well as their position, head, body orientation and F-formation information over the entire event duration. Through extensive experiments with state-of-the-art approaches, we show (a) the limitations of current methods and (b) how the recorded multiple cues synergetically aid automatic analysis of social interactions. SALSA is available at http://tev.fbk.eu/salsa
    corecore